Socially Adaptive Path Planning in Human Environments Using Inverse Reinforcement Learning
نویسندگان
چکیده
A key skill for mobile robots is the ability to navigate e ciently through their environment. In the case of social or assistive robots, this involves navigating through human crowds. Typical performance criteria, such as reaching the goal using the shortest path, are not appropriate in such environments, where it is more important for the robot to move in a socially adaptive manner such as respecting comfort zones of the pedestrians. We propose a framework for socially adaptive path planning in dynamic environments, by generating human-like path trajectory. Our framework consists of three modules: a feature extraction module, Inverse Reinforcement Learning module, and a path planning module. The feature extraction module extracts features necessary to characterize the state information, such as density and velocity of surrounding obstacles, from a RGB-Depth sensor. The Inverse Reinforcement Learning module uses a set of demonstration trajectories generated by an expert to learn the expert’s behaviour when faced with di↵erent state features, and represent it as a cost function that respects social variables. Finally, the planning module integrates a threelayer architecture, where a global path is optimized according to a classical shortest-path objective using a global map known a priori, a local path is planned over a shorter distance using the features extracted from a RGB-D sensor and the cost function inferred from Inverse Reinforcement Learning module, and a low-level Beomjoon Kim E-mail: [email protected] Joelle Pineau School of Computer Science, McGill University, 3480 University, Canada Tel.: 514-398-5432 Fax: 514-398-3883 E-mail: [email protected] system handles avoidance of immediate obstacles. We evaluate our approach by deploying it on a real robotic wheelchair platform in various scenarios, and comparing the robot trajectories to human trajectories.
منابع مشابه
Efficient Inverse Reinforcement Learning using Adaptive State-Graphs
Inverse Reinforcement Learning (IRL) provides a powerful mechanism for learning complex behaviors from demonstration by rationalizing such demonstrations. Unfortunately its applicability has been largely hindered by lack of powerful representations that can take advantage of various task affordances while still admitting scalability. Inspired by the success of sampling based approaches in class...
متن کاملPath Planning for a Robot Manipulator based on Probabilistic Roadmap and Reinforcement Learning
The probabilistic roadmap (PRM) method, which is a popular path planning scheme, for a manipulator, can find a collision-free path by connecting the start and goal poses through a roadmap constructed by drawing random nodes in the free configuration space. PRM exhibits robust performance for static environments, but its performance is poor for dynamic environments. On the other hand, reinforcem...
متن کاملGraph Based Inverse Reinforcement Learning for Social Robot Navigation
Abstract. Mobile robots that operate in human populated spaces are required to act in ways perceived as socially normative with respect to the environment considered. Past approaches have merely focussed on robot centric optimality criteria such as path lengths, heading changes, time to goal etc while ignoring social aspects like personal space intrusions. We learn how to navigate socially from...
متن کاملPath Planning of a Mobile Robot as a Discrete Optimization Problem an Adjustment of Weight Parameters in the Objective Function by Reinforcement Learning
p. 1 Champion Teams FC Portugal Team Description: RoboCup 2000 Simulation League Champion p. 29 The Cornell RoboCup Team p. 41 CS Freiburg: Doing the Right Thing in a Group p. 52 The UNSW RoboCup 2000 Sony Legged League Team p. 64 Challenge Award Finalists Adaptive Path Planner for Highly Dynamic Environments p. 76 Communication and Coordination Among Heterogeneous Mid-Size Players: ART99 p. 86...
متن کاملMotivated Learning from Interesting Events: Adaptive, Multitask Learning Agents for Complex Environments
This paper presents a model of motivation in learning agents to achieve adaptive, multi-task learning in complex, dynamic environments. Previously, computational models of motivation have been considered as speed-up or attention focus mechanisms for planning and reinforcement learning systems, however these different models do not provide a unified approach to the development or evaluation of c...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- I. J. Social Robotics
دوره 8 شماره
صفحات -
تاریخ انتشار 2016